# Causal language model
Progen2 Xlarge
Bsd-3-clause
ProGen2-xlarge is a protein generation model based on the foundational model proposed by Nijkamp et al., with modifications to its configuration and forward propagation.
Large Language Model
Transformers

P
hugohrban
38
1
Progen2 Base
Bsd-3-clause
ProGen2-base is a protein generation model based on the research by Nijkamp et al., supporting protein sequence generation and prediction.
Large Language Model
Transformers

P
hugohrban
4,937
3
Progen2 Small
Bsd-3-clause
ProGen2-small is a mirrored version of the protein generation model based on Nijkamp et al.'s research, with slight adjustments to the configuration and forward propagation process.
Large Language Model
Transformers

P
hugohrban
6,505
2
Vortex 3b
Other
vortex-3b is a 2.78 billion parameter causal language model developed by OEvortex, based on the Pythia-2.8b model and fine-tuned on the Vortex-50k dataset.
Large Language Model
Transformers English

V
OEvortex
16
5
Pythia 2.8b
Apache-2.0
Pythia - 2.8 billion is a member of the scalable language model suite developed by EleutherAI, designed specifically to promote the interpretability research of large language models. This model is based on the Transformer architecture and is trained on the The Pile dataset, with 2.8 billion parameters.
Large Language Model
Transformers English

P
EleutherAI
40.38k
30
Gpt2 Base Thai
MIT
Thai causal language model based on GPT-2 architecture, trained on the OSCAR dataset
Large Language Model Other
G
flax-community
1,026
10
Tool
A GPT-2 language model trained on Tool band lyrics, capable of generating stylistically similar lyric texts
Large Language Model English
T
huggingartists
14
1
The Notorious Big
A GPT-2 language model fine-tuned on lyrics by The Notorious B.I.G., capable of generating text in a similar lyrical style
Large Language Model English
T
huggingartists
16
0
Gpt Neo 1.3B Vietnamese News
This is a Vietnamese causal language model with 1.3B parameters trained on Vietnamese news data
Large Language Model
Transformers Other

G
VietAI
990
28
Rugpt3medium Sum Gazeta
Apache-2.0
Russian abstractive summarization model based on rugpt3medium_based_on_gpt2, specifically trained on the Gazeta dataset
Text Generation
Transformers Other

R
IlyaGusev
1,228
4
Gpt Neo 1.3B
MIT
GPT-Neo 1.3B is a 1.3 billion parameter autoregressive language model developed by EleutherAI, based on the GPT-3 architecture, excelling in text generation tasks
Large Language Model English
G
EleutherAI
208.93k
298
Featured Recommended AI Models